19 research outputs found
Identifiability and consistent estimation of nonparametric translation hidden Markov models with general state space
This paper considers hidden Markov models where the observations are given as
the sum of a latent state which lies in a general state space and some
independent noise with unknown distribution. It is shown that these fully
nonparametric translation models are identifiable with respect to both the
distribution of the latent variables and the distribution of the noise, under
mostly a light tail assumption on the latent variables. Two nonparametric
estimation methods are proposed and we prove that the corresponding estimators
are consistent for the weak convergence topology. These results are illustrated
with numerical experiments
Support and distribution inference from noisy data
We consider noisy observations of a distribution with unknown support. In the
deconvolution model, it has been proved recently [19] that, under very mild
assumptions, it is possible to solve the deconvolution problem without knowing
the noise distribution and with no sample of the noise. We first give general
settings where the theory applies and provide classes of supports that can be
recovered in this context. We then exhibit classes of distributions over which
we prove adaptive minimax rates (up to a log log factor) for the estimation of
the support in Hausdorff distance. Moreover, for the class of distributions
with compact support, we provide estimators of the unknown (in general
singular) distribution and prove maximum rates in Wasserstein distance. We also
prove an almost matching lower bound on the associated minimax risk
Contrôle non-asymptotique de l'EMV pour les modèles de Markov cachés mal spécifiés
We study the problem of estimating an unknown time process distribution using nonparametric hidden Markov models in the misspecified setting, that is when the true distribution of the process may not come from a hidden Markov model. We show that when the true distribution is exponentially mixing and satisfies a forgetting assumption, the maximum likelihood estimator recovers the best approximation of the true distribution. We prove a finite sample bound on the resulting error and show that it is optimal in the minimax sense–up to logarithmic factors–when the model is well specified
Consistent order estimation for nonparametric Hidden Markov Models
We consider the problem of estimating the number of hidden states (the order) of a nonparametric hidden Markov model (HMM). We propose two different methods and prove their almost sure consistency without any prior assumption, be it on the order or on the emission distributions. This is the first time a consistency result is proved in such a general setting without using restrictive assumptions such as a priori upper bounds on the order or parametric restrictions on the emission distributions. Our main method relies on the minimization of a penalized least squares criterion. In addition to the consistency of the order estimation, we also prove that this method yields rate minimax adaptive estimators of the parameters of the HMM - up to a logarithmic factor. Our second method relies on estimating the rank of a matrix obtained from the distribution of two consecutive observations. Finally, numerical experiments are used to compare both methods and study their ability to select the right order in several situations
Multi-agent learning via gradient ascent activity-based credit assignment
Abstract We consider the situation in which cooperating agents learn to achieve a common goal based solely on a global return that results from all agents’ behavior. The method proposed is based on taking into account the agents’ activity, which can be any additional information to help solving multi-agent decentralized learning problems. We propose a gradient ascent algorithm and assess its performance on synthetic data
On the convergence of the MLE as an estimator of the learning rate in the Exp3 algorithm
When fitting the learning data of an individual to algorithm-like learning
models, the observations are so dependent and non-stationary that one may
wonder what the classical Maximum Likelihood Estimator (MLE) could do, even if
it is the usual tool applied to experimental cognition. Our objective in this
work is to show that the estimation of the learning rate cannot be efficient if
the learning rate is constant in the classical Exp3 (Exponential weights for
Exploration and Exploitation) algorithm. Secondly, we show that if the learning
rate decreases polynomially with the sample size, then the prediction error and
in some cases the estimation error of the MLE satisfy bounds in probability
that decrease at a polynomial rate
Support and distribution inference from noisy data
We consider noisy observations of a distribution with unknown support. In the deconvolution model, it has been proved recently [19] that, under very mild assumptions, it ispossible to solve the deconvolution problem without knowing the noise distribution and with no sample of the noise. We first give general settings where the theory applies and provide classes of supports that can be recovered in this context. We then exhibit classes of distributions over which we prove adaptive minimax rates (up to a log log factor) for the estimation of the support in Hausdorff distance. Moreover, for the class of distributions with compact support, we provide estimators of the unknown (in general singular) distribution and prove maximum rates in Wasserstein distance. We also prove an almost matching lower bound on the associated minimax risk